Goto

Collaborating Authors

 Prince William Sound


MH-GIN: Multi-scale Heterogeneous Graph-based Imputation Network for AIS Data (Extended Version)

Liu, Hengyu, Li, Tianyi, He, Yuqiang, Torp, Kristian, Li, Yushuai, Jensen, Christian S.

arXiv.org Artificial Intelligence

Location-tracking data from the Automatic Identification System, much of which is publicly available, plays a key role in a range of maritime safety and monitoring applications. However, the data suffers from missing values that hamper downstream applications. Imputing the missing values is challenging because the values of different heterogeneous attributes are updated at diverse rates, resulting in the occurrence of multi-scale dependencies among attributes. Existing imputation methods that assume similar update rates across attributes are unable to capture and exploit such dependencies, limiting their imputation accuracy. We propose MH-GIN, a Multi-scale Heterogeneous Graph-based Imputation Network that aims improve imputation accuracy by capturing multi-scale dependencies. Specifically, MH-GIN first extracts multi-scale temporal features for each attribute while preserving their intrinsic heterogeneous characteristics. Then, it constructs a multi-scale heterogeneous graph to explicitly model dependencies between heterogeneous attributes to enable more accurate imputation of missing values through graph propagation. Experimental results on two real-world datasets find that MH-GIN is capable of an average 57% reduction in imputation errors compared to state-of-the-art methods, while maintaining computational efficiency. The source code and implementation details of MH-GIN are publicly available https://github.com/hyLiu1994/MH-GIN.


MetaKP: On-Demand Keyphrase Generation

Wu, Di, Shen, Xiaoxian, Chang, Kai-Wei

arXiv.org Artificial Intelligence

Traditional keyphrase prediction methods predict a single set of keyphrases per document, failing to cater to the diverse needs of users and downstream applications. To bridge the gap, we introduce on-demand keyphrase generation, a novel paradigm that requires keyphrases that conform to specific high-level goals or intents. For this task, we present MetaKP, a large-scale benchmark comprising four datasets, 7500 documents, and 3760 goals across news and biomedical domains with human-annotated keyphrases. Leveraging MetaKP, we design both supervised and unsupervised methods, including a multi-task fine-tuning approach and a self-consistency prompting method with large language models. The results highlight the challenges of supervised fine-tuning, whose performance is not robust to distribution shifts. By contrast, the proposed self-consistency prompting approach greatly improves the performance of large language models, enabling GPT-4o to achieve 0.548 SemF1, surpassing the performance of a fully fine-tuned BART-base model. Finally, we demonstrate the potential of our method to serve as a general NLP infrastructure, exemplified by its application in epidemic event detection from social media.


RS5M and GeoRSCLIP: A Large Scale Vision-Language Dataset and A Large Vision-Language Model for Remote Sensing

Zhang, Zilun, Zhao, Tiancheng, Guo, Yulong, Yin, Jianwei

arXiv.org Artificial Intelligence

Pre-trained Vision-Language Models (VLMs) utilizing extensive image-text paired data have demonstrated unprecedented image-text association capabilities, achieving remarkable results across various downstream tasks. A critical challenge is how to make use of existing large-scale pre-trained VLMs, which are trained on common objects, to perform the domain-specific transfer for accomplishing domain-related downstream tasks. A critical challenge is how to make use of existing large-scale pre-trained VLMs, which are trained on common objects, to perform the domain-specific transfer for accomplishing domain-related downstream tasks. In this paper, we propose a new framework that includes the Domain pre-trained Vision-Language Model (DVLM), bridging the gap between the General Vision-Language Model (GVLM) and domain-specific downstream tasks. Moreover, we present an image-text paired dataset in the field of remote sensing (RS), RS5M, which has 5 million RS images with English descriptions. The dataset is obtained from filtering publicly available image-text paired datasets and captioning label-only RS datasets with pre-trained VLM. These constitute the first large-scale RS image-text paired dataset. Additionally, we fine-tuned the CLIP model and tried several Parameter-Efficient Fine-Tuning methods on RS5M to implement the DVLM. Experimental results show that our proposed dataset is highly effective for various tasks, and our model GeoRSCLIP improves upon the baseline or previous state-of-the-art model by $3\%\sim20\%$ in Zero-shot Classification (ZSC), $3\%\sim6\%$ in Remote Sensing Cross-Modal Text-Image Retrieval (RSCTIR) and $4\%\sim5\%$ in Semantic Localization (SeLo) tasks. Dataset and models have been released in: \url{https://github.com/om-ai-lab/RS5M}.


Artificial Intelligence and the Information Lifecycle

#artificialintelligence

The year is 1989 and we're introduced to the World Wide Web. The Berlin Wall is coming down. The Exxon Valdez is spilling oil in Prince William Sound, Alaska. Students are calling for democracy and free speech in Tiananmen Square. Crockett and Tubbs are clearing the mean streets of Miami.


What a Great Lakes shipwreck could tell us about American history

Christian Science Monitor | Science

The second-oldest confirmed shipwreck in the Great Lakes, an American-built, Canadian-owned sloop that sank in Lake Ontario more than 200 years ago, has been found, a team of underwater explorers said Wednesday. The three-member western New York-based team said it discovered the shipwreck this summer in deep water off Oswego, in central New York. Images captured by a remotely operated vehicle confirmed it is the Washington, which sank during a storm in 1803, team member Jim Kennard said. "This one is very special. We don't get too many like this," said Mr. Kennard, who along with Roger Pawlowski and Roland "Chip" Stevens has found numerous wrecks in Lake Ontario and other waterways.


Trial by Fire: Understanding the Design Requirements for Agents in Complex Environments

Cohen, Paul R., Greenberg, Michael L., Hart, David M., Howe, Adele E.

AI Magazine

Second, These sections describe how Phoenix agents there are motivating issues, of plan in real time but do not provide the which the foremost is to understand minute detail that is offered elsewhere (Cohen how complex environments et al. forthcoming). The next section illustrates constrain on the design of Phoenix agents controlling a forest fire. We seek general The last section describes the current status of rules that justify and explain the project and our immediate goals. The terms in these rules describe The Phoenix task is to control simulated characteristics of environments, forest fires by deploying simulated bulldozers, tasks and behaviors, and the crews, airplanes, and other objects. We discuss architectures of agents. Phoenix Environment, Layers 1 and 2 but Phoenix is a commentary on the Phoenix Simulator. In the following pages, we describe Phoenix from the perspective of our technical aims and motives. The second section describes the Phoenix task--controlling simulated forest fires-- and explains why we use a simulated environment instead of a real, physical one. The two lowest layers of Phoenix, described in The Phoenix Environment, Layers 1 and 2, implement the simulated environment and maintain the illusion that the forest fire and agents are acting simultaneously. Above these layers are two others: a Figure 2. Fire at 12:30 Bulldozers are Close to organization of multiple Meeting at the Fire Front. The left pane displays the real world; the right pane displays fireboss sees it. Firefighting objects are also and other agents are semiautonomous.